Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 22.664
Filtrar
1.
Sci Rep ; 14(1): 7731, 2024 04 02.
Artículo en Inglés | MEDLINE | ID: mdl-38565928

RESUMEN

Data storage in DNA has recently emerged as a promising archival solution, offering space-efficient and long-lasting digital storage solutions. Recent studies suggest leveraging the inherent redundancy of synthesis and sequencing technologies by using composite DNA alphabets. A major challenge of this approach involves the noisy inference process, obstructing large composite alphabets. This paper introduces a novel approach for DNA-based data storage, offering, in some implementations, a 6.5-fold increase in logical density over standard DNA-based storage systems, with near-zero reconstruction error. Combinatorial DNA encoding uses a set of clearly distinguishable DNA shortmers to construct large combinatorial alphabets, where each letter consists of a subset of shortmers. We formally define various combinatorial encoding schemes and investigate their theoretical properties. These include information density and reconstruction probabilities, as well as required synthesis and sequencing multiplicities. We then propose an end-to-end design for a combinatorial DNA-based data storage system, including encoding schemes, two-dimensional (2D) error correction codes, and reconstruction algorithms, under different error regimes. We performed simulations and show, for example, that the use of 2D Reed-Solomon error correction has significantly improved reconstruction rates. We validated our approach by constructing two combinatorial sequences using Gibson assembly, imitating a 4-cycle combinatorial synthesis process. We confirmed the successful reconstruction, and established the robustness of our approach for different error types. Subsampling experiments supported the important role of sampling rate and its effect on the overall performance. Our work demonstrates the potential of combinatorial shortmer encoding for DNA-based data storage and describes some theoretical research questions and technical challenges. Combining combinatorial principles with error-correcting strategies, and investing in the development of DNA synthesis technologies that efficiently support combinatorial synthesis, can pave the way to efficient, error-resilient DNA-based storage solutions.


Asunto(s)
Replicación del ADN , ADN , Análisis de Secuencia de ADN/métodos , ADN/genética , Algoritmos , Almacenamiento y Recuperación de la Información
2.
BMC Med Imaging ; 24(1): 86, 2024 Apr 10.
Artículo en Inglés | MEDLINE | ID: mdl-38600525

RESUMEN

Medical imaging AI systems and big data analytics have attracted much attention from researchers of industry and academia. The application of medical imaging AI systems and big data analytics play an important role in the technology of content based remote sensing (CBRS) development. Environmental data, information, and analysis have been produced promptly using remote sensing (RS). The method for creating a useful digital map from an image data set is called image information extraction. Image information extraction depends on target recognition (shape and color). For low-level image attributes like texture, Classifier-based Retrieval(CR) techniques are ineffective since they categorize the input images and only return images from the determined classes of RS. The issues mentioned earlier cannot be handled by the existing expertise based on a keyword/metadata remote sensing data service model. To get over these restrictions, Fuzzy Class Membership-based Image Extraction (FCMIE), a technology developed for Content-Based Remote Sensing (CBRS), is suggested. The compensation fuzzy neural network (CFNN) is used to calculate the category label and fuzzy category membership of the query image. Use a basic and balanced weighted distance metric. Feature information extraction (FIE) enhances remote sensing image processing and autonomous information retrieval of visual content based on time-frequency meaning, such as color, texture and shape attributes of images. Hierarchical nested structure and cyclic similarity measure produce faster queries when searching. The experiment's findings indicate that applying the proposed model can have favorable outcomes for assessment measures, including Ratio of Coverage, average means precision, recall, and efficiency retrieval that are attained more effectively than the existing CR model. In the areas of feature tracking, climate forecasting, background noise reduction, and simulating nonlinear functional behaviors, CFNN has a wide range of RS applications. The proposed method CFNN-FCMIE achieves a minimum range of 4-5% for all three feature vectors, sample mean and comparison precision-recall ratio, which gives better results than the existing classifier-based retrieval model. This work provides an important reference for medical imaging artificial intelligence system and big data analysis.


Asunto(s)
Inteligencia Artificial , Tecnología de Sensores Remotos , Humanos , Ciencia de los Datos , Almacenamiento y Recuperación de la Información , Redes Neurales de la Computación
3.
J Korean Med Sci ; 39(14): e127, 2024 Apr 15.
Artículo en Inglés | MEDLINE | ID: mdl-38622936

RESUMEN

BACKGROUND: To overcome the limitations of relying on data from a single institution, many researchers have studied data linkage methodologies. Data linkage includes errors owing to legal issues surrounding personal information and technical issues related to data processing. Linkage errors affect selection bias, and external and internal validity. Therefore, quality verification for each connection method with adherence to personal information protection is an important issue. This study evaluated the linkage quality of linked data and analyzed the potential bias resulting from linkage errors. METHODS: This study analyzed claims data submitted to the Health Insurance Review and Assessment Service (HIRA DATA). The linkage errors of the two deterministic linkage methods were evaluated based on the use of the match key. The first deterministic linkage uses a unique identification number, and the second deterministic linkage uses the name, gender, and date of birth as a set of partial identifiers. The linkage error included in this deterministic linkage method was compared with the absolute standardized difference (ASD) of Cohen's according to the baseline characteristics, and the linkage quality was evaluated through the following indicators: linked rate, false match rate, missed match rate, positive predictive value, sensitivity, specificity, and F1-score. RESULTS: For the deterministic linkage method that used the name, gender, and date of birth as a set of partial identifiers, the true match rate was 83.5 and the missed match rate was 16.5. Although there was bias in some characteristics of the data, most of the ASD values were less than 0.1, with no case greater than 0.5. Therefore, it is difficult to determine whether linked data constructed with deterministic linkages have substantial differences. CONCLUSION: This study confirms the possibility of building health and medical data at the national level as the first data linkage quality verification study using big data from the HIRA. Analyzing the quality of linkages is crucial for comprehending linkage errors and generating reliable analytical outcomes. Linkers should increase the reliability of linked data by providing linkage error-related information to researchers. The results of this study will serve as reference data to increase the reliability of multicenter data linkage studies.


Asunto(s)
Almacenamiento y Recuperación de la Información , Registro Médico Coordinado , Humanos , Reproducibilidad de los Resultados , Registro Médico Coordinado/métodos , Valor Predictivo de las Pruebas , Servicios de Salud
4.
Database (Oxford) ; 20242024 Apr 15.
Artículo en Inglés | MEDLINE | ID: mdl-38625809

RESUMEN

The National Health and Nutrition Examination Survey provides comprehensive data on demographics, sociology, health and nutrition. Conducted in 2-year cycles since 1999, most of its data are publicly accessible, making it pivotal for research areas like studying social determinants of health or tracking trends in health metrics such as obesity or diabetes. Assembling the data and analyzing it presents a number of technical and analytic challenges. This paper introduces the nhanesA R package, which is designed to assist researchers in data retrieval and analysis and to enable the sharing and extension of prior research efforts. We believe that fostering community-driven activity in data reproducibility and sharing of analytic methods will greatly benefit the scientific community and propel scientific advancements. Database URL: https://github.com/cjendres1/nhanes.


Asunto(s)
Almacenamiento y Recuperación de la Información , Encuestas Nutricionales , Reproducibilidad de los Resultados , Bases de Datos Factuales
5.
PLoS One ; 19(4): e0301760, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38625954

RESUMEN

Cloud computing alludes to the on-demand availability of personal computer framework resources, primarily information storage and processing power, without the customer's direct personal involvement. Cloud computing has developed dramatically among many organizations due to its benefits such as cost savings, resource pooling, broad network access, and ease of management; nonetheless, security has been a major concern. Researchers have proposed several cryptographic methods to offer cloud data security; however, their execution times are linear and longer. A Security Key 4 Optimization Algorithm (SK4OA) with a non-linear run time is proposed in this paper. The secret key of SK4OA determines the run time rather than the size of the data as such is able to transmit large volumes of data with minimal bandwidth and able to resist security attacks like brute force since its execution timings are unpredictable. A data set from Kaggle was used to determine the algorithm's mean and standard deviation after thirty (30) times of execution. Data sizes of 3KB, 5KB, 8KB, 12KB, and 16 KB were used in this study. There was an empirical analysis done against RC4, Salsa20, and Chacha20 based on encryption time, decryption time, throughput and memory utilization. The analysis showed that SK4OA generated lowest mean non-linear run time of 5.545±2.785 when 16KB of data was executed. Additionally, SK4OA's standard deviation was greater, indicating that the observed data varied far from the mean. However, RC4, Salsa20, and Chacha20 showed smaller standard deviations making them more clustered around the mean resulting in predictable run times.


Asunto(s)
Algoritmos , Almacenamiento y Recuperación de la Información , Nube Computacional , Seguridad Computacional , Microcomputadores
6.
BMC Bioinformatics ; 25(1): 152, 2024 Apr 16.
Artículo en Inglés | MEDLINE | ID: mdl-38627652

RESUMEN

BACKGROUND: Text summarization is a challenging problem in Natural Language Processing, which involves condensing the content of textual documents without losing their overall meaning and information content, In the domain of bio-medical research, summaries are critical for efficient data analysis and information retrieval. While several bio-medical text summarizers exist in the literature, they often miss out on an essential text aspect: text semantics. RESULTS: This paper proposes a novel extractive summarizer that preserves text semantics by utilizing bio-semantic models. We evaluate our approach using ROUGE on a standard dataset and compare it with three state-of-the-art summarizers. Our results show that our approach outperforms existing summarizers. CONCLUSION: The usage of semantics can improve summarizer performance and lead to better summaries. Our summarizer has the potential to aid in efficient data analysis and information retrieval in the field of biomedical research.


Asunto(s)
Algoritmos , Investigación Biomédica , Semántica , Almacenamiento y Recuperación de la Información , Procesamiento de Lenguaje Natural
7.
Nat Commun ; 15(1): 3293, 2024 Apr 17.
Artículo en Inglés | MEDLINE | ID: mdl-38632239

RESUMEN

DNA-based artificial motors have allowed the recapitulation of biological functions and the creation of new features. Here, we present a molecular robotic system that surveys molecular environments and reports spatial information in an autonomous and repeated manner. A group of molecular agents, termed 'crawlers', roam around and copy information from DNA-labeled targets, generating records that reflect their trajectories. Based on a mechanism that allows random crawling, we show that our system is capable of counting the number of subunits in example molecular complexes. Our system can also detect multivalent proximities by generating concatenated records from multiple local interactions. We demonstrate this capability by distinguishing colocalization patterns of three proteins inside fixed cells under different conditions. These mechanisms for examining molecular landscapes may serve as a basis towards creating large-scale detailed molecular interaction maps inside the cell with nanoscale resolution.


Asunto(s)
Procedimientos Quirúrgicos Robotizados , ADN , Proteínas , Fenómenos Biofísicos , Almacenamiento y Recuperación de la Información
8.
Neural Comput ; 36(5): 781-802, 2024 Apr 23.
Artículo en Inglés | MEDLINE | ID: mdl-38658027

RESUMEN

Variation in the strength of synapses can be quantified by measuring the anatomical properties of synapses. Quantifying precision of synaptic plasticity is fundamental to understanding information storage and retrieval in neural circuits. Synapses from the same axon onto the same dendrite have a common history of coactivation, making them ideal candidates for determining the precision of synaptic plasticity based on the similarity of their physical dimensions. Here, the precision and amount of information stored in synapse dimensions were quantified with Shannon information theory, expanding prior analysis that used signal detection theory (Bartol et al., 2015). The two methods were compared using dendritic spine head volumes in the middle of the stratum radiatum of hippocampal area CA1 as well-defined measures of synaptic strength. Information theory delineated the number of distinguishable synaptic strengths based on nonoverlapping bins of dendritic spine head volumes. Shannon entropy was applied to measure synaptic information storage capacity (SISC) and resulted in a lower bound of 4.1 bits and upper bound of 4.59 bits of information based on 24 distinguishable sizes. We further compared the distribution of distinguishable sizes and a uniform distribution using Kullback-Leibler divergence and discovered that there was a nearly uniform distribution of spine head volumes across the sizes, suggesting optimal use of the distinguishable values. Thus, SISC provides a new analytical measure that can be generalized to probe synaptic strengths and capacity for plasticity in different brain regions of different species and among animals raised in different conditions or during learning. How brain diseases and disorders affect the precision of synaptic plasticity can also be probed.


Asunto(s)
Teoría de la Información , Plasticidad Neuronal , Sinapsis , Animales , Sinapsis/fisiología , Plasticidad Neuronal/fisiología , Espinas Dendríticas/fisiología , Región CA1 Hipocampal/fisiología , Modelos Neurológicos , Almacenamiento y Recuperación de la Información , Masculino , Hipocampo/fisiología , Ratas
9.
J Biomed Semantics ; 15(1): 3, 2024 Apr 23.
Artículo en Inglés | MEDLINE | ID: mdl-38654304

RESUMEN

BACKGROUND: Systematic reviews of Randomized Controlled Trials (RCTs) are an important part of the evidence-based medicine paradigm. However, the creation of such systematic reviews by clinical experts is costly as well as time-consuming, and results can get quickly outdated after publication. Most RCTs are structured based on the Patient, Intervention, Comparison, Outcomes (PICO) framework and there exist many approaches which aim to extract PICO elements automatically. The automatic extraction of PICO information from RCTs has the potential to significantly speed up the creation process of systematic reviews and this way also benefit the field of evidence-based medicine. RESULTS: Previous work has addressed the extraction of PICO elements as the task of identifying relevant text spans or sentences, but without populating a structured representation of a trial. In contrast, in this work, we treat PICO elements as structured templates with slots to do justice to the complex nature of the information they represent. We present two different approaches to extract this structured information from the abstracts of RCTs. The first approach is an extractive approach based on our previous work that is extended to capture full document representations as well as by a clustering step to infer the number of instances of each template type. The second approach is a generative approach based on a seq2seq model that encodes the abstract describing the RCT and uses a decoder to infer a structured representation of a trial including its arms, treatments, endpoints and outcomes. Both approaches are evaluated with different base models on a manually annotated dataset consisting of RCT abstracts on an existing dataset comprising 211 annotated clinical trial abstracts for Type 2 Diabetes and Glaucoma. For both diseases, the extractive approach (with flan-t5-base) reached the best F 1 score, i.e. 0.547 ( ± 0.006 ) for type 2 diabetes and 0.636 ( ± 0.006 ) for glaucoma. Generally, the F 1 scores were higher for glaucoma than for type 2 diabetes and the standard deviation was higher for the generative approach. CONCLUSION: In our experiments, both approaches show promising performance extracting structured PICO information from RCTs, especially considering that most related work focuses on the far easier task of predicting less structured objects. In our experimental results, the extractive approach performs best in both cases, although the lead is greater for glaucoma than for type 2 diabetes. For future work, it remains to be investigated how the base model size affects the performance of both approaches in comparison. Although the extractive approach currently leaves more room for direct improvements, the generative approach might benefit from larger models.


Asunto(s)
Indización y Redacción de Resúmenes , Ensayos Clínicos Controlados Aleatorios como Asunto , Humanos , Procesamiento de Lenguaje Natural , Almacenamiento y Recuperación de la Información/métodos
10.
Sci Rep ; 14(1): 7147, 2024 03 26.
Artículo en Inglés | MEDLINE | ID: mdl-38532119

RESUMEN

E-health has become a top priority for healthcare organizations focused on advancing healthcare services. Thus, medical organizations have been widely adopting cloud services, resulting in the effective storage of sensitive data. To prevent privacy and security issues associated with the data, attribute-based encryption (ABE) has been a popular choice for encrypting private data. Likewise, the attribute-based access control (ABAC) technique has been widely adopted for controlling data access. Researchers have proposed electronic health record (EHR) systems using ABE techniques like ciphertext policy attribute-based encryption (CP-ABE), key policy attribute-based encryption (KP-ABE), and multi authority attribute-based encryption (MA-ABE). However, there is a lack of rigorous comparison among the various ABE schemes used in healthcare systems. To better understand the usability of ABE techniques in medical systems, we performed a comprehensive review and evaluation of the three popular ABE techniques by developing EHR systems using knowledge graphs with the same data but different encryption mechanisms. We have used the MIMIC-III dataset with varying record sizes for this study. This paper can help healthcare organizations or researchers using ABE in their systems to comprehend the correct usage scenario and the prospect of ABE deployment in the most recent technological evolution.


Asunto(s)
Registros Electrónicos de Salud , Almacenamiento y Recuperación de la Información , Algoritmos , Seguridad Computacional , Nube Computacional , Atención a la Salud
11.
Comput Methods Programs Biomed ; 248: 108110, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38452685

RESUMEN

BACKGROUND AND OBJECTIVE: High-resolution (HR) MR images provide rich structural detail to assist physicians in clinical diagnosis and treatment plan. However, it is arduous to acquire HR MRI due to equipment limitations, scanning time or patient comfort. Instead, HR MRI could be obtained through a number of computer assisted post-processing methods that have proven to be effective and reliable. This paper aims to develop a convolutional neural network (CNN) based super-resolution reconstruction framework for low-resolution (LR) T2w images. METHOD: In this paper, we propose a novel multi-modal HR MRI generation framework based on deep learning techniques. Specifically, we construct a CNN based on multi-resolution analysis to learn an end-to-end mapping between LR T2w and HR T2w, where HR T1w is fed into the network to offer detailed a priori information to help generate HR T2w. Furthermore, a low-frequency filtering module is introduced to filter out the interference from HR-T1w during high-frequency information extraction. Based on the idea of multi-resolution analysis, detailed features extracted from HR T1w and LR T2w are fused at two scales in the network and then HR T2w is reconstructed by upsampling and dense connectivity module. RESULTS: Extensive quantitative and qualitative evaluations demonstrate that the proposed method enhances the recovered HR T2w details and outperforms other state-of-the-art methods. In addition, the experimental results also suggest that our network has a lightweight structure and favorable generalization performance. CONCLUSION: The results show that the proposed method is capable of reconstructing HR T2w with higher accuracy. Meanwhile, the super-resolution reconstruction results on other dataset illustrate the excellent generalization ability of the method.


Asunto(s)
Almacenamiento y Recuperación de la Información , Médicos , Humanos , Imagen por Resonancia Magnética , Redes Neurales de la Computación , Procesamiento de Imagen Asistido por Computador
12.
IEEE Trans Nanobioscience ; 23(2): 310-318, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38546987

RESUMEN

In nanopore sequencers, single-stranded DNA molecules (or k-mers) enter a small opening in a membrane called a nanopore and modulate the ionic current through the pore, producing a channel output in the form of a noisy piecewise constant signal. An important problem in DNA-based data storage is finding a set of k-mers, i.e. a DNA code, that is robust against noisy sample duplication introduced by nanopore sequencers. Good DNA codes should contain as many k-mers as possible that produce distinguishable current signals (squiggles) as measured by the sequencer. The dissimilarity between squiggles can be estimated using a bound on their pairwise error probability, which is used as a metric for code design. Unfortunately, code construction using the union bound is limited to small k's due to the difficulty of finding maximum cliques in large graphs. In this paper, we construct large codes by concatenating codewords from a base code, thereby packing more information in a single strand while retaining the storage efficiency of the base code. To facilitate decoding, we include a circumfix in the base code to reduce the effect of the nanopore channel memory. We show that the decoding complexity scales as [Formula: see text], where m is the number of concatenated k-mers. Simulations show that the base code error rate is stable as m increases.


Asunto(s)
ADN Concatenado , Nanoporos , ADN/genética , Análisis de Secuencia de ADN , Almacenamiento y Recuperación de la Información
13.
J Cancer Res Clin Oncol ; 150(3): 140, 2024 Mar 19.
Artículo en Inglés | MEDLINE | ID: mdl-38504034

RESUMEN

PURPOSE: Despite advanced technologies in breast cancer management, challenges remain in efficiently interpreting vast clinical data for patient-specific insights. We reviewed the literature on how large language models (LLMs) such as ChatGPT might offer solutions in this field. METHODS: We searched MEDLINE for relevant studies published before December 22, 2023. Keywords included: "large language models", "LLM", "GPT", "ChatGPT", "OpenAI", and "breast". The risk bias was evaluated using the QUADAS-2 tool. RESULTS: Six studies evaluating either ChatGPT-3.5 or GPT-4, met our inclusion criteria. They explored clinical notes analysis, guideline-based question-answering, and patient management recommendations. Accuracy varied between studies, ranging from 50 to 98%. Higher accuracy was seen in structured tasks like information retrieval. Half of the studies used real patient data, adding practical clinical value. Challenges included inconsistent accuracy, dependency on the way questions are posed (prompt-dependency), and in some cases, missing critical clinical information. CONCLUSION: LLMs hold potential in breast cancer care, especially in textual information extraction and guideline-driven clinical question-answering. Yet, their inconsistent accuracy underscores the need for careful validation of these models, and the importance of ongoing supervision.


Asunto(s)
Neoplasias de la Mama , Humanos , Femenino , Neoplasias de la Mama/terapia , Mama , Almacenamiento y Recuperación de la Información , Lenguaje
14.
Comput Biol Med ; 173: 108354, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38522251

RESUMEN

Colorectal cancer (CRC) is a leading cause of cancer-related deaths, with colonic crypts (CC) being crucial in its development. Accurate segmentation of CC is essential for decisions CRC and developing diagnostic strategies. However, colonic crypts' blurred boundaries and morphological diversity bring substantial challenges for automatic segmentation. To mitigate this problem, we proposed the Dual-Branch Asymmetric Encoder-Decoder Segmentation Network (DAUNet), a novel and efficient model tailored for confocal laser endomicroscopy (CLE) CC images. In DAUNet, we crafted a dual-branch feature extraction module (DFEM), employing Focus operations and dense depth-wise separable convolution (DDSC) to extract multiscale features, boosting semantic understanding and coping with the morphological diversity of CC. We also introduced the feature fusion guided module (FFGM) to adaptively combine features from both branches using cross-group spatial and channel attention to improve the model representation in focusing on specific lesion features. These modules are seamlessly integrated into the encoder for effective multiscale information extraction and fusion, and DDSC is further introduced in the decoder to provide rich representations for precise segmentation. Moreover, the local multi-layer perceptron (LMLP) module is designed to decouple and recalibrate features through a local linear transformation that filters out the noise and refines features to provide edge-enriched representation. Experimental evaluations on two datasets demonstrate that the proposed method achieves Intersection over Union (IoU) scores of 81.54% and 84.83%, respectively, which are on par with state-of-the-art methods, exhibiting its effectiveness for CC segmentation. The proposed method holds great potential in assisting physicians with precise lesion localization and region analysis, thereby improving the diagnostic accuracy of CRC.


Asunto(s)
Colon , 60670 , Colon/diagnóstico por imagen , Almacenamiento y Recuperación de la Información , Redes Neurales de la Computación , Semántica , Procesamiento de Imagen Asistido por Computador
15.
Comput Biol Med ; 173: 108291, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38522254

RESUMEN

BACKGROUND: It is very important to detect mandibular fracture region. However, the size of mandibular fracture region is different due to different anatomical positions, different sites and different degrees of force. It is difficult to locate and recognize fracture region accurately. METHODS: To solve these problems, M3YOLOv5 model is proposed in this paper. Three feature enhancement strategies are designed, which improve the ability of model to locate and recognize mandibular fracture region. Firstly, Global-Local Feature Extraction Module (GLFEM) is designed. By effectively combining Convolutional Neural Network (CNN) and Transformer, the problem of insufficient global information extraction ability of CNN is complemented, and the positioning ability of the model to the fracture region is improved. Secondly, in order to improve the interaction ability of context information, Deep-Shallow Feature Interaction Module (DSFIM) is designed. In this module, the spatial information in the shallow feature layer is embedded to the deep feature layer by the spatial attention mechanism, and the semantic information in the deep feature layer is embedded to the shallow feature layer by the channel attention mechanism. The fracture region recognition ability of the model is improved. Finally, Multi-scale Multi receptive-field Feature Mixing Module (MMFMM) is designed. Deep separate convolution chains are used in this modal, which is composed by multiple layers of different scales and different dilation coefficients. This method provides richer receptive field for the model, and the ability to detect fracture region of different scales is improved. RESULTS: The precision rate, mAP value, recall rate and F1 value of M3YOLOv5 model on mandibular fracture CT data set are 97.18%, 96.86%, 94.42% and 95.58% respectively. The experimental results show that there is better performance about M3YOLOv5 model than the mainstream detection models. CONCLUSION: The M3YOLOv5 model can effectively recognize and locate the mandibular fracture region, which is of great significance for doctors' clinical diagnosis.


Asunto(s)
Fracturas Mandibulares , Humanos , Fracturas Mandibulares/diagnóstico por imagen , Almacenamiento y Recuperación de la Información , Redes Neurales de la Computación , Semántica
16.
Int J Med Inform ; 186: 105415, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38520907

RESUMEN

INTRODUCTION: Health records serve not only as a database of a patient's health history and treatment process but also as a crucial tool for doctors to diagnose and treat patients. However, the storage and sharing of these records are sensitive issues as they involve maintaining patient privacy and ensuring data transparency, security, and interoperability between different parties. Challenges to achieving these goals in the current surgical process can impact the allocation of medical resources and surgical outcomes. METHODS: This article proposes a healthcare 5.0 framework for medical surgery that deploys a secure and distributed network using Blockchain to demonstrate transactions between different parties in the orthopedic surgery process. The proposed network uses the Hyperledger Composer platform for deployment, and a patient-doctor-supplier orthopedic surgery network is designed and implemented to enable the safe sharing of medical records. RESULTS: A benchmarking tool was implemented for analyzing different scenarios of applying blockchain technology to orthopedic surgery. The application of blockchain technology to orthopedic surgery presents a promising solution for data sharing and supply chain management in the field. The integration of blockchain with cloud storage and hybrid encryption ensures secure and efficient storage of Electronic Health Record (EHR) and Personal Health Record (PHR) data. By leveraging the tamper-proof nature of blockchain and addressing concerns regarding centralized data storage, this scenario demonstrates enhanced security, improved access efficiency, and privacy protection in medical data sharing. CONCLUSIONS: The article demonstrates the feasibility of using an IoT-based blockchain network in orthopedic surgery, which can reduce medical errors and improve data interoperability among different parties. This unique application of blockchain enables secure sharing of medical records, ensuring transparency, security, and interoperability. The network design may also be applicable to other surgeries and medical applications in the future.


Asunto(s)
Cadena de Bloques , Humanos , Registros Electrónicos de Salud , Atención a la Salud , Confidencialidad , Almacenamiento y Recuperación de la Información , Seguridad Computacional
17.
Int J Med Inform ; 185: 105380, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38447318

RESUMEN

INTRODUCTION: Electronic health records (EHR) are of great value for clinical research. However, EHR consists primarily of unstructured text which must be analysed by a human and coded into a database before data analysis- a time-consuming and costly process limiting research efficiency. Natural language processing (NLP) can facilitate data retrieval from unstructured text. During AssistMED project, we developed a practical, NLP tool that automatically provides comprehensive clinical characteristics of patients from EHR, that is tailored to clinical researchers needs. MATERIAL AND METHODS: AssistMED retrieves patient characteristics regarding clinical conditions, medications with dosage, and echocardiographic parameters with clinically oriented data structure and provides researcher-friendly database output. We validate the algorithm performance against manual data retrieval and provide critical quantitative and qualitative analysis. RESULTS: AssistMED analysed the presence of 56 clinical conditions, medications from 16 drug groups with dosage and 15 numeric echocardiographic parameters in a sample of 400 patients hospitalized in the cardiology unit. No statistically significant differences between algorithm and human retrieval were noted. Qualitative analysis revealed that disagreements with manual annotation were primarily accounted to random algorithm errors, erroneous human annotation and lack of advanced context awareness of our tool. CONCLUSIONS: Current NLP approaches are feasible to acquire accurate and detailed patient characteristics tailored to clinical researchers' needs from EHR. We present an in-depth description of an algorithm development and validation process, discuss obstacles and pinpoint potential solutions, including opportunities arising with recent advancements in the field of NLP, such as large language models.


Asunto(s)
Cardiología , Procesamiento de Lenguaje Natural , Humanos , Registros Electrónicos de Salud , Algoritmos , Almacenamiento y Recuperación de la Información
18.
PLoS One ; 19(3): e0299506, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38489324

RESUMEN

Thorough examination of renal biopsies may improve understanding of renal disease. Imaging of renal biopsies with fluorescence nonlinear microscopy (NLM) and optical clearing enables three-dimensional (3D) visualization of pathology without microtome sectioning. Archival renal paraffin blocks from 12 patients were deparaffinized and stained with Hoechst and Eosin for fluorescent nuclear and cytoplasmic/stromal contrast, then optically cleared using benzyl alcohol benzyl benzoate (BABB). NLM images of entire biopsy fragments (thickness range 88-660 µm) were acquired using NLM with fluorescent signals mapped to an H&E color scale. Cysts, glomeruli, exudative lesions, and Kimmelstiel-Wilson nodules were segmented in 3D and their volumes, diameters, and percent composition could be obtained. The glomerular count on 3D NLM volumes was high indicating that archival blocks could be a vast tissue resource to enable larger-scale retrospective studies. Rapid optical clearing and NLM imaging enables more thorough biopsy examination and is a promising technique for analysis of archival paraffin blocks.


Asunto(s)
Colorantes , Parafina , Humanos , Estudios Retrospectivos , Microscopía Fluorescente , Biopsia , Almacenamiento y Recuperación de la Información , Imagenología Tridimensional/métodos , Microscopía Confocal
19.
Clin Orthop Relat Res ; 482(4): 578-588, 2024 Apr 01.
Artículo en Inglés | MEDLINE | ID: mdl-38517757

RESUMEN

BACKGROUND: The lay public is increasingly using ChatGPT (a large language model) as a source of medical information. Traditional search engines such as Google provide several distinct responses to each search query and indicate the source for each response, but ChatGPT provides responses in paragraph form in prose without providing the sources used, which makes it difficult or impossible to ascertain whether those sources are reliable. One practical method to infer the sources used by ChatGPT is text network analysis. By understanding how ChatGPT uses source information in relation to traditional search engines, physicians and physician organizations can better counsel patients on the use of this new tool. QUESTIONS/PURPOSES: (1) In terms of key content words, how similar are ChatGPT and Google Search responses for queries related to topics in orthopaedic surgery? (2) Does the source distribution (academic, governmental, commercial, or form of a scientific manuscript) differ for Google Search responses based on the topic's level of medical consensus, and how is this reflected in the text similarity between ChatGPT and Google Search responses? (3) Do these results vary between different versions of ChatGPT? METHODS: We evaluated three search queries relating to orthopaedic conditions: "What is the cause of carpal tunnel syndrome?," "What is the cause of tennis elbow?," and "Platelet-rich plasma for thumb arthritis?" These were selected because of their relatively high, medium, and low consensus in the medical evidence, respectively. Each question was posed to ChatGPT version 3.5 and version 4.0 20 times for a total of 120 responses. Text network analysis using term frequency-inverse document frequency (TF-IDF) was used to compare text similarity between responses from ChatGPT and Google Search. In the field of information retrieval, TF-IDF is a weighted statistical measure of the importance of a key word to a document in a collection of documents. Higher TF-IDF scores indicate greater similarity between two sources. TF-IDF scores are most often used to compare and rank the text similarity of documents. Using this type of text network analysis, text similarity between ChatGPT and Google Search can be determined by calculating and summing the TF-IDF for all keywords in a ChatGPT response and comparing it with each Google search result to assess their text similarity to each other. In this way, text similarity can be used to infer relative content similarity. To answer our first question, we characterized the text similarity between ChatGPT and Google Search responses by finding the TF-IDF scores of the ChatGPT response and each of the 20 Google Search results for each question. Using these scores, we could compare the similarity of each ChatGPT response to the Google Search results. To provide a reference point for interpreting TF-IDF values, we generated randomized text samples with the same term distribution as the Google Search results. By comparing ChatGPT TF-IDF to the random text sample, we could assess whether TF-IDF values were statistically significant from TF-IDF values obtained by random chance, and it allowed us to test whether text similarity was an appropriate quantitative statistical measure of relative content similarity. To answer our second question, we classified the Google Search results to better understand sourcing. Google Search provides 20 or more distinct sources of information, but ChatGPT gives only a single prose paragraph in response to each query. So, to answer this question, we used TF-IDF to ascertain whether the ChatGPT response was principally driven by one of four source categories: academic, government, commercial, or material that took the form of a scientific manuscript but was not peer-reviewed or indexed on a government site (such as PubMed). We then compared the TF-IDF similarity between ChatGPT responses and the source category. To answer our third research question, we repeated both analyses and compared the results when using ChatGPT 3.5 versus ChatGPT 4.0. RESULTS: The ChatGPT response was dominated by the top Google Search result. For example, for carpal tunnel syndrome, the top result was an academic website with a mean TF-IDF of 7.2. A similar result was observed for the other search topics. To provide a reference point for interpreting TF-IDF values, a randomly generated sample of text compared with Google Search would have a mean TF-IDF of 2.7 ± 1.9, controlling for text length and keyword distribution. The observed TF-IDF distribution was higher for ChatGPT responses than for random text samples, supporting the claim that keyword text similarity is a measure of relative content similarity. When comparing source distribution, the ChatGPT response was most similar to the most common source category from Google Search. For the subject where there was strong consensus (carpal tunnel syndrome), the ChatGPT response was most similar to high-quality academic sources rather than lower-quality commercial sources (TF-IDF 8.6 versus 2.2). For topics with low consensus, the ChatGPT response paralleled lower-quality commercial websites compared with higher-quality academic websites (TF-IDF 14.6 versus 0.2). ChatGPT 4.0 had higher text similarity to Google Search results than ChatGPT 3.5 (mean increase in TF-IDF similarity of 0.80 to 0.91; p < 0.001). The ChatGPT 4.0 response was still dominated by the top Google Search result and reflected the most common search category for all search topics. CONCLUSION: ChatGPT responses are similar to individual Google Search results for queries related to orthopaedic surgery, but the distribution of source information can vary substantially based on the relative level of consensus on a topic. For example, for carpal tunnel syndrome, where there is widely accepted medical consensus, ChatGPT responses had higher similarity to academic sources and therefore used those sources more. When fewer academic or government sources are available, especially in our search related to platelet-rich plasma, ChatGPT appears to have relied more heavily on a small number of nonacademic sources. These findings persisted even as ChatGPT was updated from version 3.5 to version 4.0. CLINICAL RELEVANCE: Physicians should be aware that ChatGPT and Google likely use the same sources for a specific question. The main difference is that ChatGPT can draw upon multiple sources to create one aggregate response, while Google maintains its distinctness by providing multiple results. For topics with a low consensus and therefore a low number of quality sources, there is a much higher chance that ChatGPT will use less-reliable sources, in which case physicians should take the time to educate patients on the topic or provide resources that give more reliable information. Physician organizations should make it clear when the evidence is limited so that ChatGPT can reflect the lack of quality information or evidence.


Asunto(s)
Síndrome del Túnel Carpiano , Motor de Búsqueda , Humanos , Almacenamiento y Recuperación de la Información
20.
J Forensic Sci ; 69(3): 1075-1087, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38443323

RESUMEN

iPhone operating system (iOS) devices utilize binary cookies as a data storage tool, encoding user-specific information within an often-neglected element of smartphone analysis. This binary format contains details such as cookie flags, expiration, and creation dates, domain, and value of the cookie. These data are invaluable for forensic investigations. This study presents a comprehensive methodology to decode and extract valuable data from these files, enhancing the ability to recover user activity information from iOS devices. This paper provides an in-depth forensic investigation into the structure and function of iOS binary cookie files. Our proposed forensic technique includes a combination of reverse engineering and custom-built Python scripts to decode the binary structure. The results of our research demonstrate that these cookie files can reveal an array of important digital traces, including user preferences, visited websites, and timestamps of online activities. It concludes that the forensic analysis of iOS binary cookie files can be a tool for forensic investigators and cybersecurity professionals. In the rapidly evolving domain of digital forensics, this research contributes to our understanding of less-explored data sources within iOS devices and their potential value in investigative contexts.


Asunto(s)
Ciencias Forenses , Teléfono Inteligente , Humanos , Ciencias Forenses/métodos , Aplicaciones Móviles , Almacenamiento y Recuperación de la Información , Programas Informáticos
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...